Learning Imprecise Hidden Markov Models
نویسندگان
چکیده
Consider a stationary precise hidden Markov model (HMM) with n hidden states Xk, taking values xk in a set {1, . . . ,m} and n observations Ok, taking values ok. Both the marginal model pX1(x1), the emission models pOk|Xk(ok|xk) and the transition models pXk|Xk−1(xk|xk−1) are unknown. We can then use the Baum–Welch algorithm [see, e.g., 4] to get a maximum-likelihood estimate of these models. The Baum–Welch algorithm constructs the expected number of transitions ni j := ∑k=2 pXk−1,Xk|O1,...,On(i, j|o1, . . . ,on) from state i to state j in the whole Markov chain. If we do not have enough data to justify a precise probability model, such as the one we obtain using the classical Baum–Welch algorithm, then the approach we present is useful. Our contribution exists of a method for learning imprecise transition probabilities in an HMM. We are not aware of another such method in the literature. These transitions from a state Xk−1 = i to a state Xk = j are multinomial processes. The imprecise Dirichlet model (IDM) is a convenient model for describing uncertainty about such processes [3]. In order to learn using an IDM, we need the number of transitions and a choice for the pseudocounts s. Since the hidden states are unavailable, our method consists in taking the expected1 number of transitions (positive real numbers instead of natural numbers), derived from the Baum–Welch algorithm, rather than real counts. So, the lower probability for state j conditional on state i is estimated by Q({ j}|i) := ni j/(s+∑j=1 ni j). Learning an imprecise HMM, like our method does, is necessary before being able to make imprecise inferences. There are algorithms, like MePICTIr [2] and EstiHMM [1], to make exact inferences in an imprecise HMM. We applied our method to the following application. Given an observation sequence of counted earthquakes in a number of consecutive years, we are interested in predicting the number of earthquakes in the next years. The observed number of earthquakes in a year k is assumed to be generated by a Poisson process with rate λXk . Both the emission probabilities and the marginal probability are kept precise, while the transition probabilities are made imprecise, using our method. The following figure shows the credal sets for the transition model in the case m = 3, for different pseudocounts s. λ2 λ1 λ3 s = 1 s = 2 s = 3 s = 5 s = 10 s = 20 s = 50
منابع مشابه
Partially-Hidden Markov Models
This paper addresses the problem of Hidden Markov Models (HMM) training and inference when the training data are composed of feature vectors plus uncertain and imprecise labels. The “soft” labels represent partial knowledge about the possible states at each time step and the “softness” is encoded by belief functions. For the obtained model, called a Partially-Hidden Markov Model (PHMM), the tra...
متن کاملRobust classification of multivariate time series by imprecise hidden Markov models
A novel technique to classify time series with imprecise hidden Markov models is presented. The learning of these models is achieved by coupling the EM algorithm with the imprecise Dirichlet model. In the stationarity limit, each model corresponds to an imprecise mixture of Gaussian densities, this reducing the problem to the classification of static, imprecise-probabilistic, information. Two c...
متن کاملEstiHMM: an efficient algorithm for state sequence prediction in imprecise hidden Markov models
We develop an efficient algorithm that calculates the maximal state sequences in an imprecise hidden Markov model by means of coherent lower previsions. Initial results show that this algorithm is able to robustify the inferences made by a classical precise model. Keywords— Imprecise hidden Markov model, coherent lower prevision, epistemic irrelevance, maximal state sequence, Viterbi algorithm.
متن کاملTraining Generalized Hidden Markov Model With Interval Probability Parameters
Recently generalized interval probability was proposed as a new mathematical formalism of imprecise probability. It provides a simplified probabilistic calculus based on its definitions of conditional probability and independence. The Markov property can be described in a form similar to classical probability. In this paper, an expectation-maximization approach is developed to train generalized...
متن کاملIntroducing Busy Customer Portfolio Using Hidden Markov Model
Due to the effective role of Markov models in customer relationship management (CRM), there is a lack of comprehensive literature review which contains all related literatures. In this paper the focus is on academic databases to find all the articles that had been published in 2011 and earlier. One hundred articles were identified and reviewed to find direct relevance for applying Markov models...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2011